Vector spaces, while fundamental in linear algebra, are actually a part of a larger structure known as a module.
A module is a generalization of a vector space, where the scalars come from a ring instead of a field.
To refresh your memory, a ring is a set with two operations: addition and multiplication.
A field is a ring where every nonzero element has a multiplicative inverse; i.e., you can divide by any nonzero element.
Common fields we encounter are the real numbers and the complex numbers .
Modules show up in many areas of mathematics;
A vector space is just a module over a field.
An abelian group is a module over the integers .
Recall that abelian groups are groups where the group operation is commutative, i.e., .
And a lot of other structures can be thought of as modules over some ring.
With this in mind, it is sometimes useful to think of vector spaces as modules, especially when we want to generalize the concept to other structures.
Then, instead of proving a theorem for each structure, we can prove it once for modules and apply it to all structures that are modules.
In this page, we will discuss the underlying algebraic structures that form the basis of linear algebra.
A group is a set equipped with a binary operation that satisfies a few properties:
The identity element is unique, and the inverse of an element is unique.
Groups are a fundamental concept in mathematics because groups fundamentally describe symmetry.
Imagine a group as a set of transformations that preserve some structure.
For instance, consider the group of rotations of a square. Denote a rotation by , where is the angle of rotation.
Then, the group of rotations of a square is the set . We can see that it satisfies the group axioms:
Closure: Rotating a square by and then by is the same as rotating it by .
Associativity: Rotating a square by and then by and then by is the same as rotating it by and then by and then by .
Identity: Rotating a square by doesn't change it.
Inverse: Rotating a square by and then by is the same as rotating it by .
Hopefully this sheds some light on why these axioms are defined the way they are, instead of just being arbitrary rules.
Now consider an infinite chain of beads, and let be the operation of shifting all beads one position to the right.
Then, the set of operations forms a group under the operation , where shifts all beads positions to the right.
Does this group have an identity element? Does it have inverses?
Closure: Shifting all beads one position to the right and then shifting them again is the same as shifting them two positions to the right.
Associativity: Shifting all beads one position to the right and then shifting them two positions to the right and then shifting them three positions to the right is the same as shifting them one position to the right and then shifting them two positions to the right and then shifting them three positions to the right;
Identity: Shifting all beads zero positions to the right doesn't change them.
Inverse: Shifting all beads one position to the right and then shifting them one position to the left is the same as shifting them zero positions to the right;
But wait: shifting a chain of beads... isn't that just addition?
Indeed, addition is a group operation on the integers .
The integers form a group under addition, and the nonzero integers almost form a group under multiplication.
(The notation denotes the set of elements in that are not in . It's analogus to "subtracting" the elements in from .)
We do not include in the set of nonzero integers because does not have a multiplicative inverse. That is, there's no such that .
The multiplicative inverse of an integer is , but this is not an integer for all .
Hence, the nonzero integers do not form a group under multiplication.
Subtraction does not form a group because it is non-associative. That is to say, .
Consider the set of rational numbers .
Prove that the rational numbers form a group under addition.
Define . All this is saying is that a rational number is a fraction of two integers.
Then, if we can write any number as a fraction, we know that it is a rational number.
Closure: The sum of two rational numbers is a rational number.
Associativity: The sum of three rational numbers is the same regardless of how we group them.
Identity: The identity element is .
Inverse: The inverse of a rational number is its negation.
Therefore, the rational numbers form a group under addition.
Often, we are interested in groups where the group operation is commutative.
That is, for all . This is very common in different areas of mathematics, so we give these groups a special name: abelian groups.
The name comes from the mathematician Niels Henrik Abel, who was one of the first to prove that there are no general solutions to fifth-degree polynomial equations.
(It's super fascinating to look at that proof, and one can check out the Abel-Ruffini theorem for more information.)
Consider the group of rotations of a square.
Show that this group is abelian.
(Just a side note: the terms "prove" and "show" are deliberately used differently here.
When we say "prove," we mean that we need to show that something is true using a rigorous argument.
When we say "show," we mean that we need to demonstrate something, but it doesn't have to be as rigorous; rather, it's more about understanding the concept, or giving an intuition to why something is true.
Then the proof shows how it is true. This distinction is subtle but important.)
The group of rotations of a square is , where denotes a rotation by degrees.
We can think of this group as an abelian group if we consider the rotations as addition modulo .
If you rotate by , that's equivalent to adding to the current angle.
Hence, we can shift our perspective to think of this group as the integers modulo under addition.
(When two groups are essentially the same, like this, we call them isomorphic.)
We know that addition is commutative, so the group of rotations of a square is abelian.
A group gives you a set with a single operation. A field, on the other hand, gives you a set with two operations: addition and multiplication.
Each of them have an identity element which we denote as and respectively. Importantly, the identity elements are different.
The criteria for a set to be a field are:
is an abelian group under addition.
is an abelian group under multiplication.
Multiplication distributes over addition. That is, for any , .
What these mean more concretely is that, for any :
(closure under addition).
(existence of additive inverses).
(closure under multiplication).
(existence of multiplicative inverses).
(distributive property).
These should look familiar to you, as they are the properties of the real numbers and the complex numbers .
Hence, the real numbers and the complex numbers are fields.
The integers modulo , where is a prime number. (Why does have to be prime?)
. (This is known as the rational numbers adjoined with .)
To save space, we will only prove the last two sets because the first three are well-known fields.
Integers modulo : Let .
Closure under addition: The sum of two integers modulo is an integer modulo .
Additive inverses: For any integer , .
Closure under multiplication: The product of two integers modulo is an integer modulo .
Multiplicative inverses: For the multiplicative inverse, things get a bit tricky. We want to show that for any , there exists a such that .
Since is prime, then any integer is coprime to , meaning that .
The Extended Euclidean Algorithm tells us that we can find integers such that .
Then, , so is the multiplicative inverse of modulo .
Distributive property: This is inherited from the integers.
Rational numbers adjoined with : Let .
Let's define two numbers in : and .
Closure under addition:
Additive inverses:
Closure under multiplication:
Multiplicative inverses: We can show that the multiplicative inverse of is by rationalizing the denominator.
Distributive property: This is inherited from the real numbers.
Therefore, the integers modulo and the rational numbers adjoined with are fields.
The integers modulo , , where is composite (not prime).
Integers: The integers are not a field because they do not have multiplicative inverses.
For example, there is no integer such that .
Integers modulo : If is composite, then is not a field because it does not have multiplicative inverses.
For example, in , there is no integer such that .
To see why this is the case, consider the prime factorization of .
If , then and are not coprime, so there is no multiplicative inverse for or .
Therefore, the integers and the integers modulo where is composite are not fields.
Now that we have a good understanding of groups and fields, we can finally go back into linear algebra and discuss vector spaces.
Vector spaces have two "components": an abelian group of objects called vectors, and a field of scalars.
They contain two operations alongside each other: addition of vectors and scalar multiplication.
This definition requires a few properties:
is an abelian group under addition.
is closed under scalar multiplication. That is, for any and , .
and satisfy associative and distributive properties that relate the two operations:
These properties are what make vector spaces such a powerful tool in mathematics.
Let be a vector space over the field .
Prove that the following properties hold:
The zero vector, denoted , is unique.
for all .
for all .
These properties seem very obvious, but they are actually quite miraculous when you consider the fact that they relate two different distinct sets.
For instance, the third property states that the additive inverse of the field times a vector is the same as the additive inverse of the vector.
Uniqueness of the zero vector: We need to leverage the fact that is the additive identity in . This means that for any , .
Suppose there are two zero vectors and .
Then:
Hence, given any two zero vectors, they must be the same.
Thus, the zero vector is unique.
Scalar multiplication by zero: We know that is the additive identity in .
Then:
Now that we have a good understanding of vector spaces, we can move on to linear maps, or linear transformations.
Given two vector spaces and over the same field , a linear map is a function that preserves the structure of the vector space.
Generally, in algebra, given an object, there exists a map that transforms it into another object:
Object
Map
Group
Group homomorphism
Ring
Ring homomorphism
Metric space
Continuous map
Topological space
Continuous map
Manifold
Diffeomorphism
There are many different types of objects and maps. Particularly, category theory seeks to generalize these concepts; objects are just called "objects" and maps are called "morphisms."
Since we have already discussed linear maps in the previous course, we will not delve into the intuition behind them here.
We know that linear maps have the following properties:
for all .
for all and .
These can also be written as one property:
These properties greatly simplify the study of linear maps, as they allow us to understand the map by just looking at its action on the basis vectors.
If a map is bijective, meaning that (1) it is injective (one-to-one) and (2) it is surjective (onto), then it is called an isomorphism.
Matrix multiplication: Let and be matrices.
We know that matrix multiplication is linear because it satisfies:
This is extensively discussed in the previous course.
Differential operators: Let be the differential operator, and be functions.
We know that the derivative operator is linear because it satisfies:
This is actually harder to prove than it seems; first, we need to use the formal definnition of the derivative, and then we need to use some analysis to show that limits are linear.
Integration operators: Let be the integration operator.
We know that the integration operator is linear because it satisfies:
This can be proven using the fundamental theorem of calculus. Let and be the derivatives of and , respectively.
Then:
Where the first line uses the FTOC for the entire function, and the second line uses the FTOC for the individual functions
The last two maps are not as straightforward as matrix multiplication, but they are indeed linear.
Going back to this problem, hopefully you have determined that all of the sets are vector spaces.
Show that:
is isomorphic to the set of polynomials of degree at most with coefficients in .
is isomorphic to the set of functions if .
is isomorphic to the set of polynomials: Let .
Then, we can associate with the polynomial .
We can construct a map that sends to :
To show that this map is an isomorphism, we need to show that it is bijective and linear.
Bijective: The map is clearly surjective because every polynomial of degree at most can be written as .
To show that it is injective, we need to show that if , then .
This is true because the coefficients of the polynomials are unique.
Linear: The map is linear because it satisfies:
To show this, we need to leverage the linearity of polynomial addition and scalar multiplication.
Let and .
Then:
is isomorphic to the set of functions: Let .
Then, we can associate with the function where .
To completely define , we simply need to know the values of at each element of . Siince there are elements in , this is equivalent to knowing things - which is exactly what gives us.
We can construct a map that sends to :
To show that this map is an isomorphism, we need to show that it is bijective and linear.
Bijective: The map is surjective because every function can be written as .
To show that it is injective, we need to show that if , then .
This is true because the values of the functions are unique.
Linear: The map is linear because it satisfies:
To show this, we need to leverage the linearity of function addition and scalar multiplication.
Let and .
Then:
Therefore, is isomorphic to the set of polynomials of degree at most with coefficients in , and is isomorphic to the set of functions if .
A subspace of a vector space is a subset that is itself a vector space.
That is, is closed under addition and scalar multiplication, and it contains the zero vector.
These are extremely important in linear algebra because they show up everywhere.
For instance, if you have a system of linear equations, the solutions form a subspace of the vector space of all possible solutions.
We already know the basics of subspaces, so let's move on to some less obvious aspects.
The intersection of two subspaces is also a subspace.
To show this, we need to show that the intersection is closed under addition and scalar multiplication.
Let and be subspaces of .
Then, for any and :
Closure under addition: Since and is a subspace, then .
Similarly, and is a subspace, so .
Hence, .
Closure under scalar multiplication: Since and is a subspace, then .
Similarly, and is a subspace, so .
Hence, .
Therefore, is a subspace of .
The union of two subspaces is not necessarily a subspace.
For instance, consider the -axis and the -axis in .
The union of these two subspaces is not a subspace because it is not closed under addition.
You can add some -axis vector and some -axis vector, but the result will not be on either axis.
The sum of two subspaces is also a subspace.
The sum of two subspaces and is the set of all vectors that can be written as the sum of a vector in and a vector in .
That is, .
To show that is a subspace, we need to show that it is closed under addition and scalar multiplication.
Let and .
Then, for any :
Closure under addition: Since and is a subspace, then .
Similarly, and is a subspace, so .
Hence, .
Closure under scalar multiplication: Since and is a subspace, then .
Similarly, and is a subspace, so .
Hence, .
As always, the advantage of abstraction is that we can prove general results that apply to all vector spaces, not just specific examples.
For example, consider the following subspace:
This is a subspace because it is closed under addition and scalar multiplication.
describes functions that are periodic with periods that are a factor of . For example, is in because it cycles every .
The basis of is , which is amazing because it shows that the sine functions are the building blocks of all periodic functions.
This is the basis (get it?) of Fourier series, which is a way to represent any periodic function as a sum of sines and cosines.
is also infinite-dimensional because there are infinitely many sine functions.
In this page, we discussed the fundamental concepts of linear algebra, going from the basic building blocks of groups and fields to the more complex structures of vector spaces and linear maps.
Here are the key points to remember:
A group is a set with a single operation that satisfies closure, associativity, identity, and inverses. They represent symmetry and transformations.
A field is a set with two operations, addition and multiplication, such that is an abelian group under addition and is an abelian group under multiplication.
A vector space is a set of vectors with two operations, addition and scalar multiplication, such that is an abelian group under addition and is closed under scalar multiplication.
Vector spaces are special cases of modules, which allow the scalars to come from a ring instead of just a field.
A linear map is a function that preserves the structure of the vector space. It is linear if it satisfies .
If a linear map is bijective, it is called an isomorphism.
A subspace of a vector space is a subset that is itself a vector space.
The intersection of two subspaces is a subspace, but the union of two subspaces is not necessarily a subspace.
The sum of two subspaces is a subspace.
In the next page, we will delve into the concept of linear independence and bases, which are crucial for understanding the structure of vector spaces.